A Residual Bootstrap for High-Dimensional Regression with Near Low-Rank Designs

نویسنده

  • Miles Lopes
چکیده

We study the residual bootstrap (RB) method in the context of high-dimensional linear regression. Specifically, we analyze the distributional approximation of linear contrasts c(β̂ρ − β), where β̂ρ is a ridge-regression estimator. When regression coefficients are estimated via least squares, classical results show that RB consistently approximates the laws of contrasts, provided that p n, where the design matrix is of size n × p. Up to now, relatively little work has considered how additional structure in the linear model may extend the validity of RB to the setting where p/n 1. In this setting, we propose a version of RB that resamples residuals obtained from ridge regression. Our main structural assumption on the design matrix is that it is nearly low rank — in the sense that its singular values decay according to a power-law profile. Under a few extra technical assumptions, we derive a simple criterion for ensuring that RB consistently approximates the law of a given contrast. We then specialize this result to study confidence intervals for mean response values X> i β, where X > i is the ith row of the design. More precisely, we show that conditionally on a Gaussian design with near low-rank structure, RB simultaneously approximates all of the laws X> i (β̂ρ − β), i = 1, . . . , n. This result is also notable as it imposes no sparsity assumptions on β. Furthermore, since our consistency results are formulated in terms of the Mallows (Kantorovich) metric, the existence of a limiting distribution is not required.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Estimation of (near) low-rank matrices with noise and high-dimensional scaling

We study an instance of high-dimensional inference in which the goal is to estimate a matrix Θ ∈ R12 on the basis of N noisy observations. The unknown matrix Θ is assumed to be either exactly low rank, or “near” low-rank, meaning that it can be wellapproximated by a matrix with low rank. We consider a standard M -estimator based on regularization by the nuclear or trace norm over matrices, and ...

متن کامل

ROUTE: Robust Outlier Estimation for Low Rank Matrix Recovery

In practice, even very high-dimensional data are typically sampled from low-dimensional subspaces but with intrusion of outliers and/or noises. Recovering the underlying structure and the pollution from the observations is key to understanding and processing such data. Besides properly modeling the low-rank structure of subspace, how to handle the pollution is core regarding the performance of ...

متن کامل

Residual Stresses Measurement in Hollow Samples Using Contour Method

Residual stresses are created usually undesirably during manufacturing processes, including casting, welding, metal forming, etc. Residual stresses alone or in combination with other factors can cause the destruction and fracture of components or significant decline in their service life. Therefore, it is crucial to measure the residual stresses. Contour method is a destructive testing method c...

متن کامل

A Tensor-Based Algorithm for the Optimal Model Reduction of High Dimensional Problems

We propose a method for the approximation of the solution of high-dimensional problems formulated in tensor spaces using low-rank approximation formats. The method can be seen as a perturbation of an ideal minimal residual method with a residual norm corresponding to the error in a solution norm of interest. We introduce and analyze an algorithm for the approximation of the best approximation i...

متن کامل

Can we trust the bootstrap in high-dimension?

We consider the performance of the bootstrap in high-dimensions for the setting of linear regression, where p < n but p/n is not close to zero. We consider ordinary least-squares as well as robust regression methods and adopt a minimalist performance requirement: can the bootstrap give us good confidence intervals for a single coordinate of ? (where is the true regression vector). We show throu...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2014